64 research outputs found
Hypergraph Learning with Line Expansion
Previous hypergraph expansions are solely carried out on either vertex level
or hyperedge level, thereby missing the symmetric nature of data co-occurrence,
and resulting in information loss. To address the problem, this paper treats
vertices and hyperedges equally and proposes a new hypergraph formulation named
the \emph{line expansion (LE)} for hypergraphs learning. The new expansion
bijectively induces a homogeneous structure from the hypergraph by treating
vertex-hyperedge pairs as "line nodes". By reducing the hypergraph to a simple
graph, the proposed \emph{line expansion} makes existing graph learning
algorithms compatible with the higher-order structure and has been proven as a
unifying framework for various hypergraph expansions. We evaluate the proposed
line expansion on five hypergraph datasets, the results show that our method
beats SOTA baselines by a significant margin
ManyDG: Many-domain Generalization for Healthcare Applications
The vast amount of health data has been continuously collected for each
patient, providing opportunities to support diverse healthcare predictive tasks
such as seizure detection and hospitalization prediction. Existing models are
mostly trained on other patients data and evaluated on new patients. Many of
them might suffer from poor generalizability. One key reason can be overfitting
due to the unique information related to patient identities and their data
collection environments, referred to as patient covariates in the paper. These
patient covariates usually do not contribute to predicting the targets but are
often difficult to remove. As a result, they can bias the model training
process and impede generalization. In healthcare applications, most existing
domain generalization methods assume a small number of domains. In this paper,
considering the diversity of patient covariates, we propose a new setting by
treating each patient as a separate domain (leading to many domains). We
develop a new domain generalization method ManyDG, that can scale to such
many-domain problems. Our method identifies the patient domain covariates by
mutual reconstruction and removes them via an orthogonal projection step.
Extensive experiments show that ManyDG can boost the generalization performance
on multiple real-world healthcare tasks (e.g., 3.7% Jaccard improvements on
MIMIC drug recommendation) and support realistic but challenging settings such
as insufficient data and continuous learning.Comment: The paper has been accepted by ICLR 2023, refer to
https://openreview.net/forum?id=lcSfirnflpW. We will release the data and
source codes here https://github.com/ycq091044/ManyD
Beyond Reverse KL: Generalizing Direct Preference Optimization with Diverse Divergence Constraints
The increasing capabilities of large language models (LLMs) raise
opportunities for artificial general intelligence but concurrently amplify
safety concerns, such as potential misuse of AI systems, necessitating
effective AI alignment. Reinforcement Learning from Human Feedback (RLHF) has
emerged as a promising pathway towards AI alignment but brings forth challenges
due to its complexity and dependence on a separate reward model. Direct
Preference Optimization (DPO) has been proposed as an alternative, and it
remains equivalent to RLHF under the reverse KL regularization constraint. This
paper presents -DPO, a generalized approach to DPO by incorporating diverse
divergence constraints. We show that under certain -divergences, including
Jensen-Shannon divergence, forward KL divergences and -divergences, the
complex relationship between the reward and optimal policy can also be
simplified by addressing the Karush-Kuhn-Tucker conditions. This eliminates the
need for estimating the normalizing constant in the Bradley-Terry model and
enables a tractable mapping between the reward function and the optimal policy.
Our approach optimizes LLMs to align with human preferences in a more efficient
and supervised manner under a broad set of divergence constraints. Empirically,
adopting these divergences ensures a balance between alignment performance and
generation diversity. Importantly, -DPO outperforms PPO-based methods in
divergence efficiency, and divergence constraints directly influence expected
calibration error (ECE).Comment: Preprin
MoTiAC: Multi-Objective Actor-Critics for Real-Time Bidding
Online real-time bidding (RTB) is known as a complex auction game where ad
platforms seek to consider various influential key performance indicators
(KPIs), like revenue and return on investment (ROI). The trade-off among these
competing goals needs to be balanced on a massive scale. To address the
problem, we propose a multi-objective reinforcement learning algorithm, named
MoTiAC, for the problem of bidding optimization with various goals.
Specifically, in MoTiAC, instead of using a fixed and linear combination of
multiple objectives, we compute adaptive weights overtime on the basis of how
well the current state agrees with the agent's prior. In addition, we provide
interesting properties of model updating and further prove that Pareto
optimality could be guaranteed. We demonstrate the effectiveness of our method
on a real-world commercial dataset. Experiments show that the model outperforms
all state-of-the-art baselines.Comment: 8 Pages, Extensive Experiment
Augmented Tensor Decomposition with Stochastic Optimization
Tensor decompositions are powerful tools for dimensionality reduction and
feature interpretation of multidimensional data such as signals. Existing
tensor decomposition objectives (e.g., Frobenius norm) are designed for fitting
raw data under statistical assumptions, which may not align with downstream
classification tasks. Also, real-world tensor data are usually high-ordered and
have large dimensions with millions or billions of entries. Thus, it is
expensive to decompose the whole tensor with traditional algorithms. In
practice, raw tensor data also contains redundant information while data
augmentation techniques may be used to smooth out noise in samples. This paper
addresses the above challenges by proposing augmented tensor decomposition
(ATD), which effectively incorporates data augmentations to boost downstream
classification. To reduce the memory footprint of the decomposition, we propose
a stochastic algorithm that updates the factor matrices in a batch fashion. We
evaluate ATD on multiple signal datasets. It shows comparable or better
performance (e.g., up to 15% in accuracy) over self-supervised and autoencoder
baselines with less than 5% of model parameters, achieves 0.6% ~ 1.3% accuracy
gain over other tensor-based baselines, and reduces the memory footprint by 9X
when compared to standard tensor decomposition algorithms.Comment: Fixed some typo
- …